Google upgrades its generative AI search with a 'Perspectives Preview' module, integrating firsthand sources like Reddit and forums to directly link user queries with real online discussions. This move aims to enhance efficiency in accessing trustworthy information, responding to users' preference for authentic experiences over over-SEO content, and aligning with the trend of social platforms as key information channels.....
Google has launched two AI research agents, Deep Research and Deep Research Max, based on the Gemini 3.1Pro architecture, now available for paid preview. They aim to automate complex research processes and drive AI from simple search toward long-term computation with deep reasoning. The standard version focuses on efficiency and low latency, suitable for real-time conversation; the Max version prioritizes research depth and accuracy, suitable for complex tasks.
Taobao's flash sale upgrade introduces an AI assistant for catering merchants, featuring voice search to execute complex tasks via commands, streamlining backend processes and boosting efficiency.....
Streaming platforms integrate ChatGPT, enabling users to find content via natural language instead of keyword searches. AI offers personalized recommendations through dialogue, enhancing discovery efficiency.....
Integrate all materials, let AI search and answer, and improve knowledge acquisition efficiency.
AI search, text selection translation, webpage summarization, enhancing reading and retrieval efficiency.
AI search engine designed specifically for developers, enhancing search efficiency.
A brand new site search experience to boost customer service efficiency.
Xai
$1.4
Input tokens/M
$3.5
Output tokens/M
2k
Context Length
Openai
$7.7
$30.8
200
-
Anthropic
$105
$525
Google
$2.1
$17.5
1k
$21
$0.7
$2.8
Alibaba
Moonshot
$4
$16
256
Bytedance
$0.8
$2
128
$54
$163
Tencent
$1
32
Baidu
$1.75
$14
400
Huawei
cpatonn
Llama-3.3-Nemotron-Super-49B-v1.5 is a large language model based on Meta Llama-3.3-70B-Instruct. It has enhanced inference, chat preference, and agent task capabilities through multi-stage post-training. It uses neural architecture search technology to significantly improve efficiency while maintaining high accuracy, and supports a context length of 128K tokens and multi-language processing.
nvidia
Llama-3.1-Nemotron-Ultra-253B-CPT-v1 is a large language model based on Meta Llama-3.1-405B-Instruct, supporting 128K tokens context length, optimized through Neural Architecture Search to achieve a good balance between accuracy and efficiency.
Kwaipilot
OASIS is a state-of-the-art code embedding model developed by Kwaipilot, integrating repository-level program analysis, the OASIS-instruct data synthesis algorithm, and a specialized fusion loss function, setting new benchmarks in code search efficiency and accuracy.
opensearch-project
OpenSearch's learned sparse retrieval model v2 mini version, encoding documents into sparse vectors to optimize search relevance and efficiency
A sparse retrieval model based on distillation technology, optimized for OpenSearch, supporting inference-free document encoding with improved search relevance and efficiency over V1
Coco AI is a unified search platform that integrates enterprise applications and data (such as Google Workspace, Dropbox, GitHub, etc.), providing cross-platform search and AI-driven team chat functions, aiming to improve work efficiency and collaboration experience.
DeepContext is an MCP server that provides symbol - aware semantic search capabilities for code assistants. By intelligently parsing code structure and semantic relationships, it can accurately locate relevant code snippets in large codebases and improve development efficiency.
An efficient task manager that supports multiple file formats (Markdown, JSON, YAML), provides powerful search, filtering, and organization functions, and is designed to minimize tool confusion and maximize LLM budget efficiency.
Codeix is a fast semantic code search tool designed for AI code assistants. By building a code index that can be committed to Git, it enables the search for symbols, references, and callers across codebases, improving the code location efficiency of AI agents.
DevContext is a powerful AI development context system that provides developers with project-centric continuous context awareness. It includes four memory types: short-term, long-term, situational, and semantic, supports code structure analysis and vector search, and significantly improves development efficiency.
Ticketron (tix) is an intelligent command-line tool that uses large language models (LLMs) to simplify the interaction with JIRA, supporting natural language creation and search for work orders and improving work efficiency.
An MCP server that enables AI assistants to directly search for and analyze Datadog logs, improving the efficiency of troubleshooting and log analysis through natural language queries.
Nabu and Nisaba are research prototype toolkits used to enhance the code understanding and development efficiency of LLM agents. Nabu, as an MCP server, provides code semantic search and structure analysis functions, supporting multiple programming languages; Nisaba provides workspace management capabilities for Claude Code through proxy injection and a TUI interface, helping agents manage context usage autonomously.
OneMCP is a universal MCP protocol aggregator that uniformly manages tools from multiple external MCP servers through a meta-tool architecture. It uses progressive discovery and semantic search technologies to significantly reduce token consumption of LLMs and improve tool discovery efficiency.